Evolving Spiking Neurons from Wheels to Wings

نویسندگان

  • Dario Floreano
  • Jean-Christophe Zufferey
  • Claudio Mattiussi
چکیده

We give an overview of the EPFL indoor flying project, whose goal is to evolve autonomous, adaptive, indoor micro-flyers. Indoor flight is still a challenge because it requires miniaturization, energy efficiency, and smart control. This ongoing project consists in developing an autonomous flying vision-based micro-robot, a bio-inspired controller composed of adaptive spiking neurons directly mapped into digital micro-controllers, and a method to evolve such a network without human intervention. This document describes the motivation and methodology used to reach our goal as well as the results of a number of experiments on vision-based wheeled and flying robots. 1 Issues and Challenges Flying a small aircraft in a sitting room is probably more challenging than flying in open sky because space is small and closed, there may be several obstacles of different shape and texture, and illumination may vary quite strongly within a few meters. Realising an autonomous, indoor flying robot is a formidable challenge that requires novel solutions for mechatronics, energy efficiency, and artificial intelligence. Today, there are not yet flying vehicles capable of autonomously navigating within a house. Insects are very good at flying inside a room and represent therefore a rich source of inspiration. A team in Berkeley is attempting to create a miniature flying robot modelled on the wing mechanics and dynamics of the flies [1]. However, these micro-mechatronic devices cannot yet fly and do not have sufficient payload for sensors and microelectronics required by autonomous flight. Therefore, in the first stage of our project at EPFL we aim at building micro-airplanes that can carry microelectronics, sensors, and batteries, equipped with two mechanisms that make insect so successful at flying in diverse and cluttered environments: vision and spiking neural networks. Vision is a very rich source of information about the environment and is also more energy efficient than other types of sensors used in robotics, such as active infrared sensors, sonar, and laser. Furthermore, the miniaturization trend driven by the demand for multi-media consumer electronics is bringing to the market increasingly smaller and cheaper vision devices. A commercial and fully packaged vision chip, composed of some hundreds photoreceptors, with plastic optics can weigh less than 0.4 grams. Biological vision systems deal mainly with spatial and temporal change in the image. Spatial change is given by the relationship among activation values of adjacent pixels measured at the same time. Spatial relationship is used to detect contrast, shapes, and landmarks. Temporal change is given by the relationship among activation values of the same pixels measured over time. Temporal relationship provides information about self-motion, motion of objects, and imminent collision. In biological systems, spatial and temporal information is captured and mapped into motor actions by neuronal networks with evolved architectures and time-dependent dynamics. In man-made systems, there are two major classes of artificial neuronal networks that can capture spatial and temporal information: Continuous Time Recurrent Neural Networks (CTRNN) [2] and Spiking Neural Networks (SNN) [3]. Some scientists have been trying to unveil the mechanisms of vision-guided behaviour by combining behavioural and neuro-physiological analysis with modelling and development on vision-guided mobile robots. Some of the major actors in this field include the teams led by Franceschini [4] at CNRS in Marseilles, by Buelthoff [5] at Max-Planck Institute in Tuebingen, and by Srinivasan [6] at the Australian National University in Canberra. Some of these models can be formalized in terms of CTRNN or as collections of non-linear filters. However, these methods require relatively high memory storage and computation power to handle time constants, synaptic weights, or other parameters and functions. Spiking neurons have been mainly studied and formalized within the biology-oriented community. In this project, we decided to investigate spiking neurons as candidates for our micro-systems because they communicate by binary events that can be easily mapped into digital micro-controllers. Furthermore, simple leakage and refractoriness in a spiking neuron can provide rich non-linear and time-dependent dynamics. Designing functional spiking networks is still a major challenge and there are not yet many learning algorithms that can be used to find a suitable set of synaptic connections for a desired behaviour. Therefore, in this project we use artificial evolution to discover minimal networks of spiking neurons coupled to vision sensor and actuators. From Wheels to Wings The first stage of the project consisted in assessing the feasibility of evolving networks of spiking neurons for vision-guided robots [7]. To keep things simple, we started our experiments on the miniature mobile robot Khepera equipped with a linear camera. Figure 1. Evolution of vision-based navigation with the Khepera robot. The robot was required to navigate as straight and fast as possible for 40 seconds in a rectangular arena with randomly spaced stripes on the walls (if the stripes are regularly spaced, it is relatively trivial to detect distance from walls). A fully recurrent network of 10 spiking neurons connected to the photoreceptors of the robot was genetically encoded and evolved on the physical robot. In particular, this experiment was aimed at studying whether functional behaviours can be achieved by simply evolving the connectivity among neurons, but not their synaptic weights (in other words, all existing connections are set to strength 1, with possible inhibitory neurons). Such a network and its genetic encoding require very small memory resources and computational power. The results showed that artificial evolution could reliably generate in about 20 generations robots that navigate without hitting walls. The second stage of the project consisted in developing a low-level implementation of the evolutionary spiking network in a PICTM micro-controller with few bytes of memory and a few MHz of clock speed. These micro-controllers are a suitable solution for micro-flyers because they require very little power, are extremely small and light, and include most of the circuitry required to interface sensors and actuators. Figure 2. The autonomous micro-robot Alice equipped with microcontroller, infrared active sensors, batteries, and two wheels. The implementation of a spiking neural network with 8 neurons and 8 input units, of its genetic encoding and fitness computation, and of a steady-state evolutionary algorithm took less than 35 bytes of memory storage, approximately 500 lines of assembly-code, and an update rate of 1ms, which is comparable to the update speed of biological neural networks. This was achieved by mapping neural dynamics and genetic operators directly into the architecture and functioning of the digital micro-controller without wasting even a single bit. The system was then evaluated on the Alice micro-robot, which is equipped with the same family of PIC micro-controllers, to evolve a navigation and obstacle-avoidance behaviour using the same fitness function described in [8]. It took less than 20 minutes for the robot to develop and retain smooth navigation abilities in a simple maze [9]. However, in these experiments we used active infrared sensors, instead of vision, because the vision module was not yet available for the Alice micro-robot. The third stage of the project consisted in evaluating the evolutionary spiking network for its ability to drive a vision-based blimp in a 5 by 5 meters room. In these experiments, we used the same algorithm developed in stage one for the Khepera experiments described above. The development of the autonomous indoor blimp took significant effort in order to provide it with the technology necessary to carry out evolutionary experiments. Figure 3. The evolutionary blimp in a room with randomly spaced stripes. Our blimp is equipped with two propellers for horizontal displacement, one propeller for vertical displacement, one active infrared sensor to detect altitude, a linear vision system facing forward, 6 antenna-like bumpers (not used in these experiments), a micro-controller, a Bluetooth chip for communication with a desktop computer, rechargeable batteries, and one anemometer to estimate forward speed. At this stage, the entire algorithm runs on the desktop PC, which exchanges vision data and motor commands with the blimp every 100 ms. The evolutionary blimp is asked to move forward as fast as possible for 60 seconds using only visual information (altitude control is provided by an automatic on-board routine). The fitness is proportional to the reading of the anemometer, which is mounted on the front of the robot. A preliminary set of experiments indicated that artificial evolution can generate in about 20 generations spiking controllers that drive the blimp around the room [10]. A number of experiments remain to be done with the blimp. These include an experiment where altitude control is left to the evolutionary spiking network and one using the microcontroller implementation that was tested on the Alice micro-robot. These and other experiments are under way at the moment of writing. Figure 4. A prototype of the indoor autonomous flyer. The fourth stage of the project, currently in progress, is the development of a micro airplane capable of indoor flight. A major requirement of such an airplane is to be slow enough to allow on-board and on-line vision acquisition, network update, and motor control using simple micro-controllers that require little power. Various prototypes have been developed and tested in wind tunnel [11]. The current prototype, shown in figure 4, weighs 45 grams, has an autonomy of 15 minutes when tele-operated, can fly within a room at walking speed, and is equipped with batteries, micro-controller, and a Bluetooth chip. Although this may not yet be the final model, it already has a payload of 10 grams, which is sufficient for a vision system and related microelectronics.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

From Wheels to Wings with Evolutionary Spiking Circuits

We give an overview of the EPFL indoor flying project, whose goal is to evolve neural controllers for autonomous, adaptive, indoor micro-flyers. Indoor flight is still a challenge because it requires miniaturization, energy efficiency, and control of nonlinear flight dynamics. This ongoing project consists of developing a flying, vision-based micro-robot, a bio-inspired controller composed of a...

متن کامل

Evolving Spiking Neural Networks in the GReaNs (Gene Regulatory evolving artificial Networks) Plaftorm

GReaNs (which stands for Genetic Regulatory evolving artificial Networks) is an artificial life software platform that has previously been used for modeling of evolution of gene regulatory networks able to process signals, control animats and direct multicellular development in two and three dimensions. The structure of the network in GReaNs is encoded in a linear genome, without imposing any r...

متن کامل

Evolving Networks Processing Signals with a Mixed Paradigm, Inspired by Gene Regulatory Networks and Spiking Neurons

In this paper we extend our artificial life platorm, called GReaNs (for Genetic Regulatory evolving artificial Networks) to allow evolution of spiking neural networks performign simple computational tasks. GReaNs has been previously used to model evolution of gene regulatory networks for processing signals, and also for controling the behaviour of unicallular animats and the devlopment of multi...

متن کامل

Vision-based navigation from wheels to wings

We describe an incremental approach towards the development of autonomous indoor flyers that use only vision to navigate in textured environments. In order to cope with the severe weight and energy constraints of such systems, we use spiking neural controllers that can be implemented in tiny micro-controllers and map visual information into motor commands. The network morphology is evolved by m...

متن کامل

Modeling with Spiking Neural Networks

This chapter reviews recent developments in the area of spiking neural networks (SNN) and summarizes the main contributions to this research field. We give background information about the functioning of biological neurons, discuss the most important mathematical neural models along with neural encoding techniques, learning algorithms, and applications of spiking neurons. As a specific applicat...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2002